Learning Linearly Separable Languages

نویسندگان

  • Aryeh Kontorovich
  • Corinna Cortes
  • Mehryar Mohri
چکیده

For a finite alphabet A, we define a class of embeddings of A∗ into an infinite-dimensional feature space X and show that its finitely supported hyperplanes define regular languages. This suggests a general strategy for learning regular languages from positive and negative examples. We apply this strategy to the piecewise testable languages, presenting an embedding under which these are precisely the linearly separable ones, and thus are efficiently learnable. Most of this work was done while the author was visiting the Hebrew University, Jerusalem, Israel. This work was supported in part by the IST Programme of the European Community, under the PASCAL Network of Excellence, IST-2002-506778. The research at CMU was supported in part by NSF ITR grant IIS-0205456. This publication only reflects the author’s views.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Learning Languages from Examples

In this project I explore the machine learning problem of learning formal languages from positive and negative examples. I describe a technique proposed by Kontorovich, Cortes, and Mohri in their paper titled “Learning Linearly Separable Languages.” The premise of this paper is that there exists a family of languages that are linearly separable and efficiently computable under a certain computa...

متن کامل

Kernel methods for learning languages

This paper studies a novel paradigm for learning formal languages from positive and negative examples which consists of mapping strings to an appropriate highdimensional feature space and learning a separating hyperplane in that space. Such mappings can often be represented flexibly with string kernels, with the additional benefit of computational efficiency. The paradigm inspected can thus be ...

متن کامل

Learning Languages with Rational Kernels

We present a general study of learning and linear separability with rational kernels, the sequence kernels commonly used in computational biology and natural language processing. We give a characterization of the class of all languages linearly separable with rational kernels and prove several properties of the class of languages linearly separable with a fixed rational kernel. In particular, w...

متن کامل

A Universal Kernel for Learning Regular Languages

We give a universal kernel that renders all the regular languages linearly separable. We are not able to compute this kernel efficiently and conjecture that it is intractable, but we do have an efficient ǫ-approximation. 1 Background Since the advent of Support Vector Machines (SVMs), kernel methods have flourished in machine learning theory [7]. Formally, a kernel is a positive definite functi...

متن کامل

One-Class Semi-Supervised Learning: Detecting Linearly Separable Class by its Mean

In this paper, we presented a novel semi-supervised one-class classification algorithm which assumes that class is linearly separable from other elements. We proved theoretically that class is linearly separable if and only if it is maximal by probability within the sets with the same mean. Furthermore, we presented an algorithm for identifying such linearly separable class utilizing linear pro...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006